Tree-Based Inference for Dirichlet Process Mixtures
نویسندگان
چکیده
The Dirichlet process mixture (DPM) is a widely used model for clustering and for general nonparametric Bayesian density estimation. Unfortunately, like in many statistical models, exact inference in a DPM is intractable, and approximate methods are needed to perform efficient inference. While most attention in the literature has been placed on Markov chain Monte Carlo (MCMC) [1, 2, 3], variational Bayesian (VB) [4] and collapsed variational methods [5], [6] recently introduced a novel class of approximation for DPMs based on Bayesian hierarchical clustering (BHC). These tree-based combinatorial approximations efficiently sum over exponentially many ways of partitioning the data and offer a novel lower bound on the marginal likelihood of the DPM [6]. In this paper we make the following contributions: (1) We show empirically that the BHC lower bounds are substantially tighter than the bounds given by VB [4] and by collapsed variational methods [5] on synthetic and real datasets. (2) We also show that BHC offers a more accurate predictive performance on these datasets. (3) We further improve the tree-based lower bounds with an algorithm that efficiently sums contributions from alternative trees. (4) We present a fast approximate method for BHC. Our results suggest that our combinatorial approximate inference methods and lower bounds may be useful not only in DPMs but in other models as well. Appearing in Proceedings of the 12 International Conference on Artificial Intelligence and Statistics (AISTATS) 2009, Clearwater Beach, Florida, USA. Volume 5 of JMLR: W&CP 5. Copyright 2009 by the authors.
منابع مشابه
Density Modeling and Clustering Using Dirichlet Diffusion Trees
I introduce a family of prior distributions over multivariate distributions, based on the use of a “Dirichlet diffusion tree” to generate exchangeable data sets. These priors can be viewed as generalizations of Dirichlet processes and of Dirichlet process mixtures, but unlike simple mixtures, they can capture the hierarchical structure present in many distributions, by means of the latent diffu...
متن کاملPractical Bayesian inference using mixtures of mixtures.
Discrete mixtures of normal distributions are widely used in modeling amplitude fluctuations of electrical potentials at synapses of human and other animal nervous systems. The usual framework has independent data values yj arising as yj = mu j + xn0 + j, where the means mu j come from some discrete prior G(mu) and the unknown xno + j's and observed xj, j = 1,...,n0, are Gaussian noise terms. A...
متن کاملOn the Inference of Dirichlet Mixture Priors for Protein Sequence Comparison
Dirichlet mixtures provide an elegant formalism for constructing and evaluating protein multiple sequence alignments. Their use requires the inference of Dirichlet mixture priors from curated sets of accurately aligned sequences. This article addresses two questions relevant to such inference: of how many components should a Dirichlet mixture consist, and how may a maximum-likelihood mixture be...
متن کاملSupplementary Material for Incremental Tree-Based Inference with Dependent Normalized Random Measures
In this section, we briefly review Dirichlet process mixture model (DPM) and derive its marginal likelihood.
متن کاملNested Hierarchical Dirichlet Process for Nonparametric Entity-Topic Analysis
The Hierarchical Dirichlet Process (HDP) is a Bayesian nonparametric prior for grouped data, such as collections of documents, where each group is a mixture of a set of shared mixture densities, or topics, where the number of topics is not fixed, but grows with data size. The Nested Dirichlet Process (NDP) builds on the HDP to cluster the documents, but allowing them to choose only from a set o...
متن کامل